Implicit Multisensory Associations Influence Voice Recognition
Identifieur interne : 000856 ( France/Analysis ); précédent : 000855; suivant : 000857Implicit Multisensory Associations Influence Voice Recognition
Auteurs : Katharina Von Kriegstein [Allemagne, Royaume-Uni] ; Anne-Lise Giraud [Allemagne, France]Source :
- PLoS Biology [ 1544-9173 ] ; 2006.
Abstract
Natural objects provide partially redundant information to the brain through different sensory modalities. For example, voices and faces both give information about the speech content, age, and gender of a person. Thanks to this redundancy, multimodal recognition is fast, robust, and automatic. In unimodal perception, however, only part of the information about an object is available. Here, we addressed whether, even under conditions of unimodal sensory input, crossmodal neural circuits that have been shaped by previous associative learning become activated and underpin a performance benefit. We measured brain activity with functional magnetic resonance imaging before, while, and after participants learned to associate either sensory redundant stimuli, i.e. voices and faces, or arbitrary multimodal combinations, i.e. voices and written names, ring tones, and cell phones or brand names of these cell phones. After learning, participants were better at recognizing unimodal auditory voices that had been paired with faces than those paired with written names, and association of voices with faces resulted in an increased functional coupling between voice and face areas. No such effects were observed for ring tones that had been paired with cell phones or names. These findings demonstrate that brief exposure to ecologically valid and sensory redundant stimulus pairs, such as voices and faces, induces specific multisensory associations. Consistent with predictive coding theories, associative representations become thereafter available for unimodal perception and facilitate object recognition. These data suggest that for natural objects effective predictive signals can be generated across sensory systems and proceed by optimization of functional connectivity between specialized cortical sensory modules.
Url:
DOI: 10.1371/journal.pbio.0040326
PubMed: 17002519
PubMed Central: 1570760
Affiliations:
Links toward previous steps (curation, corpus...)
- to stream Pmc, to step Corpus: 002151
- to stream Pmc, to step Curation: 002151
- to stream Pmc, to step Checkpoint: 002546
- to stream Ncbi, to step Merge: 000996
- to stream Ncbi, to step Curation: 000996
- to stream Ncbi, to step Checkpoint: 000996
- to stream Main, to step Merge: 005E09
- to stream Main, to step Curation: 005B84
- to stream Main, to step Exploration: 005B84
- to stream France, to step Extraction: 000856
Links to Exploration step
PMC:1570760Le document en format XML
<record><TEI><teiHeader><fileDesc><titleStmt><title xml:lang="en">Implicit Multisensory Associations Influence Voice Recognition</title>
<author><name sortKey="Von Kriegstein, Katharina" sort="Von Kriegstein, Katharina" uniqKey="Von Kriegstein K" first="Katharina" last="Von Kriegstein">Katharina Von Kriegstein</name>
<affiliation wicri:level="1"><nlm:aff id="aff1"> Department of Neurology, Johann Wolfgang Goethe University, Frankfurt am Main, Germany</nlm:aff>
<country xml:lang="fr">Allemagne</country>
<wicri:regionArea> Department of Neurology, Johann Wolfgang Goethe University, Frankfurt am Main</wicri:regionArea>
<wicri:noRegion>Frankfurt am Main</wicri:noRegion>
<wicri:noRegion>Frankfurt am Main</wicri:noRegion>
<wicri:noRegion>Frankfurt am Main</wicri:noRegion>
</affiliation>
<affiliation wicri:level="3"><nlm:aff id="aff2"> Wellcome Department of Imaging Neuroscience, Functional Imaging Laboratory, London, United Kingdom</nlm:aff>
<country xml:lang="fr">Royaume-Uni</country>
<wicri:regionArea> Wellcome Department of Imaging Neuroscience, Functional Imaging Laboratory, London</wicri:regionArea>
<placeName><settlement type="city">Londres</settlement>
<region type="country">Angleterre</region>
<region type="région" nuts="1">Grand Londres</region>
</placeName>
</affiliation>
</author>
<author><name sortKey="Giraud, Anne Lise" sort="Giraud, Anne Lise" uniqKey="Giraud A" first="Anne-Lise" last="Giraud">Anne-Lise Giraud</name>
<affiliation wicri:level="1"><nlm:aff id="aff1"> Department of Neurology, Johann Wolfgang Goethe University, Frankfurt am Main, Germany</nlm:aff>
<country xml:lang="fr">Allemagne</country>
<wicri:regionArea> Department of Neurology, Johann Wolfgang Goethe University, Frankfurt am Main</wicri:regionArea>
<wicri:noRegion>Frankfurt am Main</wicri:noRegion>
<wicri:noRegion>Frankfurt am Main</wicri:noRegion>
<wicri:noRegion>Frankfurt am Main</wicri:noRegion>
</affiliation>
<affiliation wicri:level="1"><nlm:aff id="aff3"> Département d'Études Cognitives, Ecole Normale Supérieure, Paris, France</nlm:aff>
<country xml:lang="fr">France</country>
<wicri:regionArea> Département d'Études Cognitives, Ecole Normale Supérieure, Paris</wicri:regionArea>
<placeName><settlement type="city">Paris</settlement>
</placeName>
</affiliation>
</author>
</titleStmt>
<publicationStmt><idno type="wicri:source">PMC</idno>
<idno type="pmid">17002519</idno>
<idno type="pmc">1570760</idno>
<idno type="url">http://www.ncbi.nlm.nih.gov/pmc/articles/PMC1570760</idno>
<idno type="RBID">PMC:1570760</idno>
<idno type="doi">10.1371/journal.pbio.0040326</idno>
<date when="2006">2006</date>
<idno type="wicri:Area/Pmc/Corpus">002151</idno>
<idno type="wicri:Area/Pmc/Curation">002151</idno>
<idno type="wicri:Area/Pmc/Checkpoint">002546</idno>
<idno type="wicri:Area/Ncbi/Merge">000996</idno>
<idno type="wicri:Area/Ncbi/Curation">000996</idno>
<idno type="wicri:Area/Ncbi/Checkpoint">000996</idno>
<idno type="wicri:doubleKey">1544-9173:2006:Von Kriegstein K:implicit:multisensory:associations</idno>
<idno type="wicri:Area/Main/Merge">005E09</idno>
<idno type="wicri:Area/Main/Curation">005B84</idno>
<idno type="wicri:Area/Main/Exploration">005B84</idno>
<idno type="wicri:Area/France/Extraction">000856</idno>
</publicationStmt>
<sourceDesc><biblStruct><analytic><title xml:lang="en" level="a" type="main">Implicit Multisensory Associations Influence Voice Recognition</title>
<author><name sortKey="Von Kriegstein, Katharina" sort="Von Kriegstein, Katharina" uniqKey="Von Kriegstein K" first="Katharina" last="Von Kriegstein">Katharina Von Kriegstein</name>
<affiliation wicri:level="1"><nlm:aff id="aff1"> Department of Neurology, Johann Wolfgang Goethe University, Frankfurt am Main, Germany</nlm:aff>
<country xml:lang="fr">Allemagne</country>
<wicri:regionArea> Department of Neurology, Johann Wolfgang Goethe University, Frankfurt am Main</wicri:regionArea>
<wicri:noRegion>Frankfurt am Main</wicri:noRegion>
<wicri:noRegion>Frankfurt am Main</wicri:noRegion>
<wicri:noRegion>Frankfurt am Main</wicri:noRegion>
</affiliation>
<affiliation wicri:level="3"><nlm:aff id="aff2"> Wellcome Department of Imaging Neuroscience, Functional Imaging Laboratory, London, United Kingdom</nlm:aff>
<country xml:lang="fr">Royaume-Uni</country>
<wicri:regionArea> Wellcome Department of Imaging Neuroscience, Functional Imaging Laboratory, London</wicri:regionArea>
<placeName><settlement type="city">Londres</settlement>
<region type="country">Angleterre</region>
<region type="région" nuts="1">Grand Londres</region>
</placeName>
</affiliation>
</author>
<author><name sortKey="Giraud, Anne Lise" sort="Giraud, Anne Lise" uniqKey="Giraud A" first="Anne-Lise" last="Giraud">Anne-Lise Giraud</name>
<affiliation wicri:level="1"><nlm:aff id="aff1"> Department of Neurology, Johann Wolfgang Goethe University, Frankfurt am Main, Germany</nlm:aff>
<country xml:lang="fr">Allemagne</country>
<wicri:regionArea> Department of Neurology, Johann Wolfgang Goethe University, Frankfurt am Main</wicri:regionArea>
<wicri:noRegion>Frankfurt am Main</wicri:noRegion>
<wicri:noRegion>Frankfurt am Main</wicri:noRegion>
<wicri:noRegion>Frankfurt am Main</wicri:noRegion>
</affiliation>
<affiliation wicri:level="1"><nlm:aff id="aff3"> Département d'Études Cognitives, Ecole Normale Supérieure, Paris, France</nlm:aff>
<country xml:lang="fr">France</country>
<wicri:regionArea> Département d'Études Cognitives, Ecole Normale Supérieure, Paris</wicri:regionArea>
<placeName><settlement type="city">Paris</settlement>
</placeName>
</affiliation>
</author>
</analytic>
<series><title level="j">PLoS Biology</title>
<idno type="ISSN">1544-9173</idno>
<idno type="eISSN">1545-7885</idno>
<imprint><date when="2006">2006</date>
</imprint>
</series>
</biblStruct>
</sourceDesc>
</fileDesc>
<profileDesc><textClass></textClass>
</profileDesc>
</teiHeader>
<front><div type="abstract" xml:lang="en"><p>Natural objects provide partially redundant information to the brain through different sensory modalities. For example, voices and faces both give information about the speech content, age, and gender of a person. Thanks to this redundancy, multimodal recognition is fast, robust, and automatic. In unimodal perception, however, only part of the information about an object is available. Here, we addressed whether, even under conditions of unimodal sensory input, crossmodal neural circuits that have been shaped by previous associative learning become activated and underpin a performance benefit. We measured brain activity with functional magnetic resonance imaging before, while, and after participants learned to associate either sensory redundant stimuli, i.e. voices and faces, or arbitrary multimodal combinations, i.e. voices and written names, ring tones, and cell phones or brand names of these cell phones. After learning, participants were better at recognizing unimodal auditory voices that had been paired with faces than those paired with written names, and association of voices with faces resulted in an increased functional coupling between voice and face areas. No such effects were observed for ring tones that had been paired with cell phones or names. These findings demonstrate that brief exposure to ecologically valid and sensory redundant stimulus pairs, such as voices and faces, induces specific multisensory associations. Consistent with predictive coding theories, associative representations become thereafter available for unimodal perception and facilitate object recognition. These data suggest that for natural objects effective predictive signals can be generated across sensory systems and proceed by optimization of functional connectivity between specialized cortical sensory modules.</p>
</div>
</front>
<back><div1 type="bibliography"><listBibl><biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
<biblStruct></biblStruct>
</listBibl>
</div1>
</back>
</TEI>
<affiliations><list><country><li>Allemagne</li>
<li>France</li>
<li>Royaume-Uni</li>
</country>
<region><li>Angleterre</li>
<li>Grand Londres</li>
</region>
<settlement><li>Londres</li>
<li>Paris</li>
</settlement>
</list>
<tree><country name="Allemagne"><noRegion><name sortKey="Von Kriegstein, Katharina" sort="Von Kriegstein, Katharina" uniqKey="Von Kriegstein K" first="Katharina" last="Von Kriegstein">Katharina Von Kriegstein</name>
</noRegion>
<name sortKey="Giraud, Anne Lise" sort="Giraud, Anne Lise" uniqKey="Giraud A" first="Anne-Lise" last="Giraud">Anne-Lise Giraud</name>
</country>
<country name="Royaume-Uni"><region name="Angleterre"><name sortKey="Von Kriegstein, Katharina" sort="Von Kriegstein, Katharina" uniqKey="Von Kriegstein K" first="Katharina" last="Von Kriegstein">Katharina Von Kriegstein</name>
</region>
</country>
<country name="France"><noRegion><name sortKey="Giraud, Anne Lise" sort="Giraud, Anne Lise" uniqKey="Giraud A" first="Anne-Lise" last="Giraud">Anne-Lise Giraud</name>
</noRegion>
</country>
</tree>
</affiliations>
</record>
Pour manipuler ce document sous Unix (Dilib)
EXPLOR_STEP=$WICRI_ROOT/Ticri/CIDE/explor/HapticV1/Data/France/Analysis
HfdSelect -h $EXPLOR_STEP/biblio.hfd -nk 000856 | SxmlIndent | more
Ou
HfdSelect -h $EXPLOR_AREA/Data/France/Analysis/biblio.hfd -nk 000856 | SxmlIndent | more
Pour mettre un lien sur cette page dans le réseau Wicri
{{Explor lien |wiki= Ticri/CIDE |area= HapticV1 |flux= France |étape= Analysis |type= RBID |clé= PMC:1570760 |texte= Implicit Multisensory Associations Influence Voice Recognition }}
Pour générer des pages wiki
HfdIndexSelect -h $EXPLOR_AREA/Data/France/Analysis/RBID.i -Sk "pubmed:17002519" \ | HfdSelect -Kh $EXPLOR_AREA/Data/France/Analysis/biblio.hfd \ | NlmPubMed2Wicri -a HapticV1
This area was generated with Dilib version V0.6.23. |